When and how convolutional neural networks generalize to out-of-distribution category–viewpoint combinations
نویسندگان
چکیده
Object recognition and viewpoint estimation lie at the heart of visual understanding. Recent studies have suggested that convolutional neural networks (CNNs) fail to generalize out-of-distribution (OOD) category–viewpoint combinations, is, combinations not seen during training. Here we investigate when how such OOD generalization may be possible by evaluating CNNs trained classify both object category three-dimensional on identifying mechanisms facilitate generalization. We show increasing number in-distribution (data diversity) substantially improves even with same amount training data. compare learning in separate shared network architectures, observe starkly different trends while are helpful distribution, significantly outperform ones combinations. Finally, demonstrate is facilitated mechanism specialization, emergence two types neuron—neurons selective invariant viewpoint, vice versa. The combination essential for However, often were authors impact data diversity architectural choices capability generalizing
منابع مشابه
How intelligent are convolutional neural networks?
Motivated by the Gestalt pattern theory in psychology and visual perception, and the Winograd Challenge for language understanding, we design synthetic experiments to investigate a deep learning algorithm’s ability to infer simple (at least for human) semantic visual concepts, such as symmetry, counting, and uniformity, etc., from examples. A visual concept is represented by randomly generated,...
متن کاملConvolutional neural networks that teach microscopes how to image
Deep learning algorithms offer a powerful means to automatically analyze the content of medical images. However, many biological samples of interest are primarily transparent to visible light and contain features that are difficult to resolve with a standard optical microscope. Here, we use a convolutional neural network (CNN) not only to classify images, but also to optimize the physical layou...
متن کاملPeople ignore token frequency when deciding how widely to generalize
Many theoretical accounts of generalization suggest that with increasing data, people should tighten their generalizations. However, these accounts presume that the additional data points are all distinct. Other accounts, such as the adaptor grammar framework in linguistics (Johnson, Griffiths, & Goldwater, 2007), suggest that when the additional data points are identical, generalizations about...
متن کاملIntroduction to Convolutional Neural Networks
6 The convolution layer 11 6.1 What is convolution? . . . . . . . . . . . . . . . . . . . . . . . . . 11 6.2 Why to convolve? . . . . . . . . . . . . . . . . . . . . . . . . . . . 13 6.3 Convolution as matrix product . . . . . . . . . . . . . . . . . . . 15 6.4 The Kronecker product . . . . . . . . . . . . . . . . . . . . . . . 17 6.5 Backward propagation: update the parameters . . . . . . . . ...
متن کاملOut-distribution training confers robustness to deep neural networks
The easiness at which adversarial instances can be generated in deep neural networks raises some fundamental questions on their functioning and concerns on their use in critical systems. In this paper, we draw a connection between overgeneralization and adversaries: a possible cause of adversaries lies in models designed to make decisions all over the input space, leading to inappropriate highc...
متن کاملذخیره در منابع من
با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید
ژورنال
عنوان ژورنال: Nature Machine Intelligence
سال: 2022
ISSN: ['2522-5839']
DOI: https://doi.org/10.1038/s42256-021-00437-5